Observables and Operators
In quantum mechanics, we are interested in the properties of particles, such as their position, momentum, energy, and so on.
In the previous page, we dealt with the abstract state of a particle, represented by a state vector
Table of Contents
Observables
In quantum mechanics, an observable is a physical quantity that can be measured.
These are properties of a particle that we can observe, such as position
Suppose we have an observable
When we measure the energy, suppose we get a value
For all energies, then, each energy
Properties of Observables
First, observables have real eigenvalues. This is to prevent observables (like energy) from being imaginary, which is not physically meaningful.
Second, eigenvectors must span the entire space. This ensures that any state vector can be expressed as a linear combination of the eigenvectors. If the eigenvectors do not span the space, then there are states that cannot be represented by the observable. But you cannot have a particle with "none" energy or "none" momentum; it must have some value. Hence, the eigenvectors must span the space.
Third, eigenvectors must be orthogonal. This ensures that the eigenvectors are independent of each other. If not, then one eigenvector can be expressed as a linear combination of the others, which means that the observable is a superposition, not a definite state, contradicting our principles.
In the future, we will show that these properties make observables something known as Hermitian operators.
Born Rule
When we measure an observable, we get a definite value, not a superposition of values.
The probability of getting a particular value
Suppose we have a state vector
Assuming
To get the component of
One guess is that the probability of measuring
But this is not quite correct.
Another guess would be that the probability is the square of the magnitude of the component.
When we drag the state vector along a circle, the sum of squares of the components is always
Flaws in the Guess
Let's lay down what we know so far, and what we need from the probability
-
The state vector
can be expressed as a linear combination of the basis vectors and : -
The probability of measuring
is guessed to be: -
All probabilities must sum to
:
The problem with the guess is that it fails the sum condition under a change of coordinates. Suppose we have the following setup:
In this setup, we have an orthonormal basis
The coefficients correctly sum to
Next, let
Under this new basis, the state vector
These coefficients do not sum to
(Once again, using the square of the magnitudes of the components would work because the norm of the state vector is invariant under a change of basis.)
How do we Solve this?
We have shown that using the magnitudes of the components as probabilities does not work because the sum of probabilities does not remain
Our goal is to find a function
Focusing on the second equation, let's zoom in on one coefficient
In order for the sum to equal
If
This means that the sum of squares of the other coefficients must be
The key is that
Another key intuition is that the probability of measuring
This means that the function
Going back to the first equation (
Recall that
Next, take the derivative of both sides with respect to
In the first term, because
Applying the chain rule to the second term:
Rearranging to isolate
Of course, this is a general result for any component
Since both equations have equal right-hand sides, the left-hand sides must be equal as well:
This is similar to how separation of variables works when solving a partial differential equation.
We have an equation of the form
More explicitly, we can see this by differentiating both sides of
This means that the derivative of the left-hand side with respect to
Since this applies to any
Rearranging:
Finally, integrating both sides gives us the function
The important takeaway is that the probability function
Solving for the Constants
First, consider what
Plugging this into our function (
Thus
Next, recall that the sum of probabilities must be
Also remember that the sum of squares of the components must be
We have already shown that
This means that
By convention, we take
Finally, the coefficient can be found by taking the inner product of the state vector with the eigenvector. Plugging this into the equation gives us the Born rule:
Born Rule: The probability of measuring an observable with eigenvalue
The above derivation is a heuristic way to understand the Born rule. Gleason's theorem is a mathematical result that shows that the Born rule is the only way to assign probabilities to the outcomes of a quantum measurement.
Hermitian Adjoint and Hermitian Operators
Suppose we have an inner product
Recall that there is a fundamental dual correspondence between vectors and linear functionals, or kets and bras.
The question is, is there a correspondence between the operator
This other operator,
Properties of Hermitian Adjoint
-
Applying the Hermitian adjoint twice gives the original operator:
To prove this, consider the inner product
. As we know, this is equal to . We can swap the two vectors (and take the complex conjugate) to get . Then, by the definition of the Hermitian adjoint again, this is equal to . Finally, applying the conjugate symmetry again, it equals . This must equal our original expression of , so . -
The adjoint of a sum of operators is the sum of the adjoints:
To prove this, consider the inner product
:- By linearity of the inner product, this is equal to
. Then, applying the Hermitian adjoint to each term yields . Putting the terms back together, this is equal to . - By the definition of the Hermitian adjoint, this is also equal to
.
Both expressions are equal, so
. - By linearity of the inner product, this is equal to
-
The adjoint of a product of operators is the product of the adjoints in reverse order (I will call this the product rule):
To prove this, consider the inner product
:- By the definition of the Hermitian adjoint, this is equal to
. - But we can also apply it to each operator separately:
.
Both expressions are equal, so
. - By the definition of the Hermitian adjoint, this is equal to
-
The adjoint of a scalar is its complex conjugate:
This is also quite easy to prove:
- By the linearity of the right-side and the conjugate-symmetry of the inner product:
. - By the definition of the Hermitian adjoint, this is equal to
.
Thus
. - By the linearity of the right-side and the conjugate-symmetry of the inner product:
-
The adjoint of an operator "flips" its input and output spaces. This is a bit more abstract, but it is a key property of the Hermitian adjoint. Another way to put it is: if an operator
is defined to be , then its adjoint is defined to be .To see why this is the case, consider the inner product
, and suppose and . In order for the inner product to be defined, both vectors must be in the same space. Since has to be in , then must be in as well. Thus, must act on , an element in to produce a vector in . Thus is defined as .Next, using the definition of the Hermitian adjoint,
. Now, must be in because is in . Thus, must act on to produce a vector in .This is why the adjoint of an operator "flips" its input and output spaces.
More generally, the Hermitian adjoint is the operator that is used for the dual correspondence. We shall now consider applying the Hermitian adjoint to a ket vector. But wait - adjoints are applied onto operators, not vectors. What does it mean to take the adjoint of a vector?
For now, we can ignore this issue and just apply the adjoint to a vector anyway.
To see what this yields, consider taking the adjoint of an inner product:
- Since the inner product is a scalar, its adjoint is just its complex conjugate:
. And since the inner product is conjugate symmetric, this is just . - By the product rule of the adjoint, we can swap the order of the vectors and take the adjoint of each:
.
Thus we have
Of course, this is a bit hand-wavy, but it gives us an intuition behind why the adjoint of a vector is its corresponding bra. The appendix contains a more rigorous proof of this fact.
Observables and Hermitian Operators
Recall that an observable is represented by an operator
- The eigenvalues of an observable are real (since they represent measurable quantities). In other words,
. - The eigenvectors of an observable form a complete basis (since it should be possible to measure any value of the observable). In other words,
. - The eigenvectors of an observable are orthogonal (since they represent distinct outcomes). In other words,
, where is the Kronecker delta.
Imagine an observable
Since
Next, the value for
Rearranging and using the properties of Dirac notation yields:
This means that the operator
In the continuous case, such as the eigenbasis of the position operator, the sum becomes an integral:
For both cases, we can see what happens when we take the Hermitian adjoint of the operator. First, since the adjoint of a sum is the sum of adjoints:
By the product rule of the adjoint, this is equal to:
Since the adjoint of a bra/ket is the corresponding ket/bra, and the adjoint of a scalar is its complex conjugate, this simplifies to:
Since eigenvalues must be real, the complex conjugate of
Operators that are equal to their Hermitian adjoints are known as Hermitian operators.
The term
Because
Expectation Value of an Observable
The expectation value of an observable
In order to compute it, recall that the expectation value for any probability density function
In quantum mechanics, the expectation value of an observable
It is technically a postulate, but we can see why this is the case. We will do it in a continuous case, but it also applies to the discrete case.
Recall the completeness relation for the eigenvectors of a continuous observable:
We can freely insert this relation into any expression. Let's put it into the right-hand side of the expectation value:
Next, we can insert another completeness relation for the left-hand side (
Now, we have a
There is a delta function in the middle,
By conjugate symmetry,
Recall from the Born rule that
Summary and Next Steps
In this chapter, we have introduced the concept of observables in quantum mechanics. Observables are quantities that can be measured, and they are represented by Hermitian operators.
Here are the key points to remember:
-
Observables are quantities that can be measured in quantum mechanics. They correspond to "real-life" or physical quantities.
-
Their eigenvectors form a basis representing the possible outcomes of the observable, and their eigenvalues are the values of the observable.
-
After making a measurement, the state vector
collapses to the eigenvector corresponding to the measured eigenvalue. -
By intuition, observables have the following three properties:
- The eigenvalues of an observable are real numbers.
- The eigenvectors of an observable form a complete basis.
- The eigenvectors of an observable are orthogonal.
-
The operators can be written as a sum of their eigenvectors and eigenvalues:
In the continuous case, this becomes an integral:
-
The operator
is known as the projector onto the eigenvector . It is sometimes denoted as . -
The Born rule states that the probability of measuring an observable with eigenvalue
is . It followed intuitively from a few key insights about the nature of observables and the conditions for probabilities. Namely, they must sum to in all bases, and they must preserve the norm of the state vector in all bases. -
The adjoint of an operator is the operator that is used for the dual correspondence between vectors and linear functionals. The adjoint of
is denoted as , and is defined such that . -
Operators that are equal to their Hermitian adjoints are known as Hermitian operators.
-
Hermitian adjoints have the following properties:
- Applying the Hermitian adjoint twice gives the original operator.
- The adjoint of a sum of operators is the sum of the adjoints.
- The adjoint of a product of operators is the product of the adjoints in reverse order.
- The adjoint of a scalar is its complex conjugate.
- The adjoint of an operator "flips" its input and output spaces.
-
Operators representing observables are Hermitian operators.
-
The expectation value of an observable
in a state is the average value of the observable when measured many times. It is denoted as and is given by .
In the next chapter, we will take a brief detour to discuss Poisson brackets in classical mechanics, and then link them to commutators in quantum mechanics, from which we will derive the uncertainty principle. Afterwards, we will discuss how operators can be represented as matrices, and how they can be diagonalized to find their eigenvectors and eigenvalues.
References
- Quantum Sense, "Maths of Quantum Mechanics", a Youtube Playlist.
- J.J. Sakurai, "Modern Quantum Mechanics", sections 1.2-1.4.
- This post on Math Stack Exchange.
Appendix: Review of Eigenvalues and Eigenvectors
An eigenvector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The scalar factor is known as the eigenvalue corresponding to that eigenvector.
Let
Rearranging, we get:
where
This equation is known as the characteristic equation of the matrix
There is a shortcut to finding the eigenvalues of a matrix.
If
This comes from the fact that the trace is the sum of the eigenvalues and the determinant is the product of the eigenvalues (one can see this geometrically by considering the area of the parallelogram formed by the eigenvectors).
Appendix: Why is the Hermitian Adjoint of a Ket its Corresponding Bra?
This proof is a bit more abstract, and it comes from this post on Math Stack Exchange. It borrows idaes from functional analysis and the dual correspondence between vectors and linear functionals.
Recall that Hermitian adjoints act on operators, not vectors.
It would thus seem like we need to somehow interpret a vector as a linear transformation.
This comes from our dual correspondence, where a vector can be thought of as a linear functional.
In fact, let's define this formally - for a vector
With this definition,
Now that we have an idea of what it means to interpret a vector as a linear transformation, we can see what it means to take the adjoint of said transformation.
Since this transformation is
Since